confident example
Appendix for " CS-Isolate: Extracting Hard Confident Examples by Content and Style Isolation " Y exiong Lin 1 Y u Y ao
We denote observed variables with gray color and latent variables with white color. Firstly, we introduce the concept of an uncontrolled style factor . Why do confident examples encourage content-style isolation? Calculate the loss using Eq. 1 and update networks; Output: The inference networks and classifier heads q It's essential to understand that although data augmentation cannot control all style factors, it still offers the benefit of "partial isolation". This approach, therefore, ensures that styles changes don't affect the derived content representation Calculate the loss using Eq. 2 and update networks; Output: The inference networks and classifier heads q Finally, confident and unlabeled examples are used to train the models based on the MixMatch algorithm.
- North America > United States (0.05)
- Asia > China > Hong Kong (0.04)
- Asia > China > Hong Kong (0.04)
- North America > United States > Illinois > Cook County > Chicago (0.04)
- Europe > Netherlands > North Holland > Amsterdam (0.04)
CS-Isolate: Extracting Hard Confident Examples by Content and Style Isolation
Label noise widely exists in large-scale image datasets. To mitigate the side effects of label noise, state-of-the-art methods focus on selecting confident examples by leveraging semi-supervised learning. Existing research shows that the ability to extract hard confident examples, which are close to the decision boundary, significantly influences the generalization ability of the learned classifier.In this paper, we find that a key reason for some hard examples being close to the decision boundary is due to the entanglement of style factors with content factors. The hard examples become more discriminative when we focus solely on content factors, such as semantic information, while ignoring style factors. Nonetheless, given only noisy data, content factors are not directly observed and have to be inferred.To tackle the problem of inferring content factors for classification when learning with noisy labels, our objective is to ensure that the content factors of all examples in the same underlying clean class remain unchanged as their style information changes.To achieve this, we utilize different data augmentation techniques to alter the styles while regularizing content factors based on some confident examples. By training existing methods with our inferred content factors, CS-Isolate proves their effectiveness in learning hard examples on benchmark datasets. The implementation is available at https://github.com/tmllab/2023
Understanding and Improving Early Stopping for Learning with Noisy Labels
The memorization effect of deep neural network (DNN) plays a pivotal role in many state-of-the-art label-noise learning methods. To exploit this property, the early stopping trick, which stops the optimization at the early stage of training, is usually adopted. Current methods generally decide the early stopping point by considering a DNN as a whole.
Appendix for " CS-Isolate: Extracting Hard Confident Examples by Content and Style Isolation " Y exiong Lin 1 Y u Y ao
We denote observed variables with gray color and latent variables with white color. Firstly, we introduce the concept of an uncontrolled style factor . Why do confident examples encourage content-style isolation? Calculate the loss using Eq. 1 and update networks; Output: The inference networks and classifier heads q It's essential to understand that although data augmentation cannot control all style factors, it still offers the benefit of "partial isolation". This approach, therefore, ensures that styles changes don't affect the derived content representation Calculate the loss using Eq. 2 and update networks; Output: The inference networks and classifier heads q Finally, confident and unlabeled examples are used to train the models based on the MixMatch algorithm.
- North America > United States (0.05)
- Asia > China > Hong Kong (0.04)
- Asia > China > Hong Kong (0.04)
- North America > United States > Illinois > Cook County > Chicago (0.04)
- Europe > Netherlands > North Holland > Amsterdam (0.04)
Understanding and Improving Early Stopping for Learning with Noisy Labels
The memorization effect of deep neural network (DNN) plays a pivotal role in many state-of-the-art label-noise learning methods. To exploit this property, the early stopping trick, which stops the optimization at the early stage of training, is usually adopted. Current methods generally decide the early stopping point by considering a DNN as a whole.
Making Binary Classification from Multiple Unlabeled Datasets Almost Free of Supervision
Wu, Yuhao, Xia, Xiaobo, Yu, Jun, Han, Bo, Niu, Gang, Sugiyama, Masashi, Liu, Tongliang
Training a classifier exploiting a huge amount of supervised data is expensive or even prohibited in a situation, where the labeling cost is high. The remarkable progress in working with weaker forms of supervision is binary classification from multiple unlabeled datasets which requires the knowledge of exact class priors for all unlabeled datasets. However, the availability of class priors is restrictive in many real-world scenarios. To address this issue, we propose to solve a new problem setting, i.e., binary classification from multiple unlabeled datasets with only one pairwise numerical relationship of class priors (MU-OPPO), which knows the relative order (which unlabeled dataset has a higher proportion of positive examples) of two class-prior probabilities for two datasets among multiple unlabeled datasets. In MU-OPPO, we do not need the class priors for all unlabeled datasets, but we only require that there exists a pair of unlabeled datasets for which we know which unlabeled dataset has a larger class prior. Clearly, this form of supervision is easier to be obtained, which can make labeling costs almost free. We propose a novel framework to handle the MU-OPPO problem, which consists of four sequential modules: (i) pseudo label assignment; (ii) confident example collection; (iii) class prior estimation; (iv) classifier training with estimated class priors. Theoretically, we analyze the gap between estimated class priors and true class priors under the proposed framework. Empirically, we confirm the superiority of our framework with comprehensive experiments. Experimental results demonstrate that our framework brings smaller estimation errors of class priors and better performance of binary classification.
- Europe > Slovenia > Drava > Municipality of Benedikt > Benedikt (0.04)
- North America > United States > Louisiana > Orleans Parish > New Orleans (0.04)
- Asia > Japan > Honshū > Kantō > Tokyo Metropolis Prefecture > Tokyo (0.04)
- Asia > China > Hong Kong (0.04)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Uncertainty > Bayesian Inference (0.66)
- Information Technology > Artificial Intelligence > Machine Learning > Learning Graphical Models > Directed Networks > Bayesian Learning (0.66)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.46)
Self-PU: Self Boosted and Calibrated Positive-Unlabeled Training
Chen, Xuxi, Chen, Wuyang, Chen, Tianlong, Yuan, Ye, Gong, Chen, Chen, Kewei, Wang, Zhangyang
Many real-world applications have to tackle the Positive-Unlabeled (PU) learning problem, i.e., learning binary classifiers from a large amount of unlabeled data and a few labeled positive examples. While current state-of-the-art methods employ importance reweighting to design various risk estimators, they ignored the learning capability of the model itself, which could have provided reliable supervision. This motivates us to propose a novel Self-PU learning framework, which seamlessly integrates PU learning and self-training. Self-PU highlights three "self"-oriented building blocks: a self-paced training algorithm that adaptively discovers and augments confident positive/negative examples as the training proceeds; a self-calibrated instance-aware loss; and a self-distillation scheme that introduces teacher-students learning as an effective regularization for PU learning. We demonstrate the state-of-the-art performance of Self-PU on common PU learning benchmarks (MNIST and CIFAR-10), which compare favorably against the latest competitors. Moreover, we study a real-world application of PU learning, i.e., classifying brain images of Alzheimer's Disease. Self-PU obtains significantly improved results on the renowned Alzheimer's Disease Neuroimaging Initiative (ADNI) database over existing methods. The code is publicly available at: https://github.com/TAMU-VITA/Self-PU.
- Europe > Austria > Vienna (0.14)
- North America > United States > Texas > Brazos County > College Station (0.04)
- Asia > China > Jiangsu Province > Nanjing (0.04)
- Health & Medicine > Therapeutic Area > Neurology > Alzheimer's Disease (1.00)
- Education (1.00)
Learning with Confident Examples: Rank Pruning for Robust Classification with Noisy Labels
Northcutt, Curtis G., Wu, Tailin, Chuang, Isaac L.
Noisy PN learning is the problem of binary classification when training examples may be mislabeled (flipped) uniformly with noise rate rho1 for positive examples and rho0 for negative examples. We propose Rank Pruning (RP) to solve noisy PN learning and the open problem of estimating the noise rates, i.e. the fraction of wrong positive and negative labels. Unlike prior solutions, RP is time-efficient and general, requiring O(T) for any unrestricted choice of probabilistic classifier with T fitting time. We prove RP has consistent noise estimation and equivalent expected risk as learning with uncorrupted labels in ideal conditions, and derive closed-form solutions when conditions are non-ideal. RP achieves state-of-the-art noise estimation and F1, error, and AUC-PR for both MNIST and CIFAR datasets, regardless of the amount of noise and performs similarly impressively when a large portion of training examples are noise drawn from a third distribution. To highlight, RP with a CNN classifier can predict if an MNIST digit is a "one"or "not" with only 0.25% error, and 0.46 error across all digits, even when 50% of positive examples are mislabeled and 50% of observed positive labels are mislabeled negative examples.
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.14)
- North America > United States > California > San Francisco County > San Francisco (0.14)
- North America > United States > New York > New York County > New York City (0.04)
- (2 more...)